-
Notifications
You must be signed in to change notification settings - Fork 639
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Fix example int8_inference_huggingface.py #414
Fix example int8_inference_huggingface.py #414
Conversation
decapoda-research is extremely out of date. Use a newer convert. |
This issue has been automatically marked as stale because it has not had recent activity. If you think this still needs to be addressed please comment on this thread. |
I think we should change the example model used instead of fixed the issue with the tokenizers. @younesbelkada any recommendation which model we should use? |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks for fixing ! I left one comment
Co-authored-by: Younes Belkada <[email protected]>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks a lot for your contribution !
4b232ed
into
bitsandbytes-foundation:main
Thanks a lot @alexrs for your work and @younesbelkada for your work. Sorry that this took a while to merge :) |
What
When running the example
int8_inference_huggingface.py
I was getting the error:when using the latest transformers=4.29.2. According to this issue:
(See comment)
Fix
Use
LlamaForCausalLM
andLlamaTokenizer
instead ofAuto...
classes works as expected.Test plan
python examples/int8_inference_huggingface.py